Goto

Collaborating Authors

 closure operator


Effect-driven interpretation: Functors for natural language composition

Bumford, Dylan, Charlow, Simon

arXiv.org Artificial Intelligence

Computer programs are often factored into pure components -- simple, total functions from inputs to outputs -- and components that may have side effects -- errors, changes to memory, parallel threads, abortion of the current loop, etc. We make the case that human languages are similarly organized around the give and pull of pure values and impure processes, and we'll aim to show how denotational techniques from computer science can be leveraged to support elegant and illuminating analyses of natural language composition.


Partial Rankings of Optimizers

Rodemann, Julian, Blocher, Hannah

arXiv.org Machine Learning

We introduce a framework for benchmarking optimizers according to multiple criteria over various test functions. Based on a recently introduced union-free generic depth function for partial orders/rankings, it fully exploits the ordinal information and allows for incomparability. Our method describes the distribution of all partial orders/rankings, avoiding the notorious shortcomings of aggregation. This permits to identify test functions that produce central or outlying rankings of optimizers and to assess the quality of benchmarking suites. Despite its importance for machine learning research, there is no broad agreement on how to compare optimization algorithms on benchmark suites with regard to multiple criteria, see Hansen et al. (2022) for instance. This is particularly relevant for multi-objective optimization, which has diverse applications ranging from reinforcement learning (Basaklar et al., 2023; Zhu et al., 2023) to representation learning (Gu et al., 2023), neural architecture search (Lu et al., 2019) and large language models (Zhou et al., 2023). But such comparisons also arise when single-objective optimizers are evaluated with respect to several metrics, see Sivaprasad et al. (2020); Mattson et al. (2020); Dahl et al. (2023). A popular example is the duality of fixed-budget (performance) and fixed-target (speed) evaluation of deep learning optimizers, see e.g. In this work, we propose a novel framework for comparing optimizers with respect to multiple criteria over a benchmarking suite of test functions.


A note on the connectedness property of union-free generic sets of partial orders

Schollmeyer, Georg, Blocher, Hannah

arXiv.org Artificial Intelligence

This short note describes and proves a connectedness property which was introduced in Blocher et al. [2023] in the context of data depth functions for partial orders. The connectedness property gives a structural insight into union-free generic sets. These sets, presented in Blocher et al. [2023], are defined by using a closure operator on the set of all partial orders which naturally appears within the theory of formal concept analysis. In the language of formal concept analysis, the property of connectedness can be vividly proven. However, since within Blocher et al. [2023] we did not discuss formal concept analysis, we outsourced the proof to this note.


Depth Functions for Partial Orders with a Descriptive Analysis of Machine Learning Algorithms

Blocher, Hannah, Schollmeyer, Georg, Jansen, Christoph, Nalenz, Malte

arXiv.org Artificial Intelligence

We propose a framework for descriptively analyzing sets of partial orders based on the concept of depth functions. Despite intensive studies of depth functions in linear and metric spaces, there is very little discussion on depth functions for non-standard data types such as partial orders. We introduce an adaptation of the well-known simplicial depth to the set of all partial orders, the union-free generic (ufg) depth. Moreover, we utilize our ufg depth for a comparison of machine learning algorithms based on multidimensional performance measures. Concretely, we analyze the distribution of different classifier performances over a sample of standard benchmark data sets. Our results promisingly demonstrate that our approach differs substantially from existing benchmarking approaches and, therefore, adds a new perspective to the vivid debate on the comparison of classifiers.


A theory of desirable things

De Bock, Jasper

arXiv.org Artificial Intelligence

The theory of imprecise probabilities [1, 2] is often thought of as a theory of partially specified probabilities, which involves manipulating sets of probabilities and their lower and upper expectations. Its mathematical underpinnings, however, are provided by an underlying theory of sets of desirable gambles [2, 3, 4, 5, 6]: sets of gambles--rewards with an uncertain payoff--that a subject finds desirable, in the sense that she prefers those gambles to the status quo--to the trivial gamble with zero payoff. Rewards are typically taken to be expressed in units of some linear utility scale, and this them implies that positive linear combinations of desirable gambles are desirable themselves. Sets of desirable gambles that satisfy this condition (as well as some other, less essential conditions) are called coherent. Due to the geometric nature of the coherence conditions, inference with desirable gambles is typically simple and intuitive, a feature that is particularly handy, also when it comes to designing proofs. Most crucially, however, well known imprecise probability models such as credal sets (closed convex sets of probabilites), lower and upper expectations (or previsions), partial preference oderings, belief functions and lower and upper probabilities, all correspond to special cases of coherent sets of desirable gambles [4], which explains the importance of the latter as a basis for impreciseprobabilistic reasoning.


Nonlinear desirability theory

Miranda, Enrique, Zaffalon, Marco

arXiv.org Artificial Intelligence

Desirability can be understood as an extension of Anscombe and Aumann's Bayesian decision theory to sets of expected utilities. At the core of desirability lies an assumption of linearity of the scale in which rewards are measured. It is a traditional assumption used to derive the expected utility model, which clashes with a general representation of rational decision making, though. Allais has, in particular, pointed this out in 1953 with his famous paradox. We note that the utility scale plays the role of a closure operator when we regard desirability as a logical theory. This observation enables us to extend desirability to the nonlinear case by letting the utility scale be represented via a general closure operator. The new theory directly expresses rewards in actual nonlinear currency (money), much in Savage's spirit, while arguably weakening the founding assumptions to a minimum. We characterise the main properties of the new theory both from the perspective of sets of gambles and of their lower and upper prices (previsions). We show how Allais paradox finds a solution in the new theory, and discuss the role of sets of probabilities in the theory.


Delta-Closure Structure for Studying Data Distribution

Buzmakov, Aleksey, Makhalova, Tatiana, Kuznetsov, Sergei O., Napoli, Amedeo

arXiv.org Artificial Intelligence

In this paper, we revisit pattern mining and study the distribution underlying a binary dataset thanks to the closure structure which is based on passkeys, i.e., minimum generators in equivalence classes robust to noise. We introduce $\Delta$-closedness, a generalization of the closure operator, where $\Delta$ measures how a closed set differs from its upper neighbors in the partial order induced by closure. A $\Delta$-class of equivalence includes minimum and maximum elements and allows us to characterize the distribution underlying the data. Moreover, the set of $\Delta$-classes of equivalence can be partitioned into the so-called $\Delta$-closure structure. In particular, a $\Delta$-class of equivalence with a high level demonstrates correlations among many attributes, which are supported by more observations when $\Delta$ is large. In the experiments, we study the $\Delta$-closure structure of several real-world datasets and show that this structure is very stable for large $\Delta$ and does not substantially depend on the data sampling used for the analysis.


Closure operators: Complexity and applications to classification and decision-making

Bajgiran, Hamed Hamze, Echenique, Federico

arXiv.org Artificial Intelligence

We study the complexity of closure operators, with applications to machine learning and decision theory. In machine learning, closure operators emerge naturally in data classification and clustering. In decision theory, they can model equivalence of choice menus, and therefore situations with a preference for flexibility. Our contribution is to formulate a notion of complexity of closure operators, which translate into the complexity of a classifier in ML, or of a utility function in decision theory.

  Country:
  Genre: Research Report (0.64)

Learning Weakly Convex Sets in Metric Spaces

Stadtländer, Eike, Horváth, Tamás, Wrobel, Stefan

arXiv.org Artificial Intelligence

We introduce the notion of weak convexity in metric spaces, a generalization of ordinary convexity commonly used in machine learning. It is shown that weakly convex sets can be characterized by a closure operator and have a unique decomposition into a set of pairwise disjoint connected blocks. We give two generic efficient algorithms, an extensional and an intensional one for learning weakly convex concepts and study their formal properties. Our experimental results concerning vertex classification clearly demonstrate the excellent predictive performance of the extensional algorithm. Two non-trivial applications of the intensional algorithm to polynomial PAC-learnability are presented. The first one deals with learning $k$-convex Boolean functions, which are already known to be efficiently PAC-learnable. It is shown how to derive this positive result in a fairly easy way by the generic intensional algorithm. The second one is concerned with the Euclidean space equipped with the Manhattan distance. For this metric space, weakly convex sets are a union of pairwise disjoint axis-aligned hyperrectangles. We show that a weakly convex set that is consistent with a set of examples and contains a minimum number of hyperrectangles can be found in polynomial time. In contrast, this problem is known to be NP-complete if the hyperrectangles may be overlapping.


Finite Confluences and Closed Pattern Mining

Soldano, Henry

arXiv.org Artificial Intelligence

The purpose of this article is to propose and investigate a partial order structure weaker than the lattice structure and which have nice properties regarding closure operators. We extend accordingly closed pattern mining and formal concept analysis to such structures we further call confluences. The primary motivation for investigating these structures is that it allows to reduce a lattice to a part whose elements are connected, as in some graph, still preserving a useful characterization of closure operators. Our investigation also considers how reducing one of the lattice involved in a Galois connection affects the structure of the closure operators ranges. When extending this way formal concept analysis we will focus on the intensional space, i.e. in reducing the pattern language, while recent investigations rather explored the reduction of the extensional space to connected elements.